1,260 research outputs found

    A Non-Parametric Bayesian Approach to Spike Sorting

    Get PDF
    Abstract — In this work we present and apply infinite Gaussian mixture modeling, a non-parametric Bayesian method, to the problem of spike sorting. As this approach is Bayesian, it allows us to integrate prior knowledge about the problem in a principled way. Because it is non-parametric we are able to avoid model selection, a difficult problem that most current spike sorting methods do not address. We compare this approach to using penalized log likelihood to select the best from multiple finite mixture models trained by expectation maximization. We show favorable offline sorting results on real data and discuss ways to extend our model to online applications

    Detecting Human-Object Contact in Images

    Get PDF
    Humans constantly contact objects to move and perform tasks. Thus, detecting human-object contact is important for building human-centered artificial intelligence. However, there exists no robust method to detect contact between the body and the scene from an image, and there exists no dataset to learn such a detector. We fill this gap with HOT ("Human-Object conTact"), a new dataset of human-object contacts for images. To build HOT, we use two data sources: (1) We use the PROX dataset of 3D human meshes moving in 3D scenes, and automatically annotate 2D image areas for contact via 3D mesh proximity and projection. (2) We use the V-COCO, HAKE and Watch-n-Patch datasets, and ask trained annotators to draw polygons for the 2D image areas where contact takes place. We also annotate the involved body part of the human body. We use our HOT dataset to train a new contact detector, which takes a single color image as input, and outputs 2D contact heatmaps as well as the body-part labels that are in contact. This is a new and challenging task that extends current foot-ground or hand-object contact detectors to the full generality of the whole body. The detector uses a part-attention branch to guide contact estimation through the context of the surrounding body parts and scene. We evaluate our detector extensively, and quantitative results show that our model outperforms baselines, and that all components contribute to better performance. Results on images from an online repository show reasonable detections and generalizability

    ECON: Explicit Clothed humans Optimized via Normal integration

    Get PDF
    The combination of artist-curated scans, and deep implicit functions (IF), is enabling the creation of detailed, clothed, 3D humans from images. However, existing methods are far from perfect. IF-based methods recover free-form geometry but produce disembodied limbs or degenerate shapes for unseen poses or clothes. To increase robustness for these cases, existing work uses an explicit parametric body model to constrain surface reconstruction, but this limits the recovery of free-form surfaces such as loose clothing that deviates from the body. What we want is a method that combines the best properties of implicit and explicit methods. To this end, we make two key observations:(1) current networks are better at inferring detailed 2D maps than full-3D surfaces, and (2) a parametric model can be seen as a “canvas” for stitching together detailed surface patches. ECON infers high-fidelity 3D humans even in loose clothes and challenging poses, while having realistic faces and fingers. This goes beyond previous methods. Quantitative, evaluation of the CAPE and Renderpeople datasets shows that ECON is more accurate than the state of the art. Perceptual studies also show that ECON’s perceived realism is better by a large margin

    3D Human Pose Estimation via Intuitive Physics

    Get PDF
    Estimating 3D humans from images often produces implausible bodies that lean, float, or penetrate the floor. Such methods ignore the fact that bodies are typically supported by the scene. A physics engine can be used to enforce physical plausibility, but these are not differentiable, rely on unrealistic proxy bodies, and are difficult to integrate into existing optimization and learning frameworks. In contrast, we exploit novel intuitive-physics (IP) terms that can be inferred from a 3D SMPL body interacting with the scene. Inspired by biomechanics, we infer the pressure heatmap on the body, the Center of Pressure (CoP) from the heatmap, and the SMPL body's Center of Mass (CoM). With these, we develop IPMAN, to estimate a 3D body from a color image in a "stable" configuration by encouraging plausible floor contact and overlapping CoP and CoM. Our IP terms are intuitive, easy to implement, fast to compute, differentiable, and can be integrated into existing optimization and regression methods. We evaluate IPMAN on standard datasets and MoYo, a new dataset with synchronized multi-view images, ground-truth 3D bodies with complex poses, body-floor contact, CoM and pressure. IPMAN produces more plausible results than the state of the art, improving accuracy for static poses, while not hurting dynamic ones. Code and data are available for research at https://ipman.is.tue.mpg.de

    ARCTIC: A Dataset for Dexterous Bimanual Hand-Object Manipulation

    Get PDF
    Humans intuitively understand that inanimate objects do not move by themselves, but that state changes are typically caused by human manipulation (e.g., the opening of a book). This is not yet the case for machines. In part this is because there exist no datasets with ground-truth 3D annotations for the study of physically consistent and synchronised motion of hands and articulated objects. To this end, we introduce ARCTIC -- a dataset of two hands that dexterously manipulate objects, containing 2.1M video frames paired with accurate 3D hand and object meshes and detailed, dynamic contact information. It contains bi-manual articulation of objects such as scissors or laptops, where hand poses and object states evolve jointly in time. We propose two novel articulated hand-object interaction tasks: (1) Consistent motion reconstruction: Given a monocular video, the goal is to reconstruct two hands and articulated objects in 3D, so that their motions are spatio-temporally consistent. (2) Interaction field estimation: Dense relative hand-object distances must be estimated from images. We introduce two baselines ArcticNet and InterField, respectively and evaluate them qualitatively and quantitatively on ARCTIC

    Studying Kaon-pion S-wave scattering in K-matrix formalism

    Full text link
    We generalize our previous work on \pi\pi scattering to K\pi scattering, and re-analyze the experiment data of K\pi scattering below 1.6 GeV. Without any free parameter, we explain K\pi I=3/2 S-wave phase shift very well by using t-channel rho and u-channel K^* meson exchange. With the t-channel and u-channel meson exchange fixed as the background term, we fit the K\pi I=1/2 S-wave data of the LASS experiment quite well by introducing one or two s-channel resonances. It is found that there is only one s-channel resonance between K\pi threshold and 1.6 GeV, i.e., K_0^*(1430) with a mass around 1438~1486 MeV and a width about 346 MeV, while the t-channel rho exchange gives a pole at (450-480i) MeV for the amplitude.Comment: REVTeX4 file, 11 pages and 3 figure

    Reconstructing Signing Avatars from Video Using Linguistic Priors

    Get PDF
    Sign language (SL) is the primary method of communication for the 70 million Deaf people around the world. Video dictionaries of isolated signs are a core SL learning tool. Replacing these with 3D avatars can aid learning and enable AR/VR applications, improving access to technology and online media. However, little work has attempted to estimate expressive 3D avatars from SL video; occlusion, noise, and motion blur make this task difficult. We address this by introducing novel linguistic priors that are universally applicable to SL and provide constraints on 3D hand pose that help resolve ambiguities within isolated signs. Our method, SGNify, captures fine-grained hand pose, facial expression, and body movement fully automatically from in-the-wild monocular SL videos. We evaluate SGNify quantitatively by using a commercial motion-capture system to compute 3D avatars synchronized with monocular video. SGNify outperforms state-of-the-art 3D body-pose- and shape-estimation methods on SL videos. A perceptual study shows that SGNify's 3D reconstructions are significantly more comprehensible and natural than those of previous methods and are on par with the source videos. Code and data are available at sgnify.is.tue.mpg.de

    Conformational Free-Energy Landscapes for a Peptide in Saline Environments

    Get PDF
    AbstractThe conformations that proteins adopt in solution are a function of both their primary structure and surrounding aqueous environment. Recent experimental and computational work on small peptides, e.g., polyK, polyE, and polyR, have highlighted an interesting and unusual behavior in the presence of aqueous ions such as ClO4−, Na+, and K+. Notwithstanding the aforementioned studies, as of this writing, the nature of the driving force induced by the presence of ions and its role on the conformational stability of peptides remains only partially understood. Molecular-dynamics simulations have been performed on the heptapeptide AEAAAEA in NaCl and KCl solutions at concentrations of 0.5, 1.0, and 2.0 M. Metadynamics in conjunction with a three-dimensional model reaction coordinate was used to sample the conformational space of the peptide. All simulations were run for 2 ÎŒs. Free-energy landscapes were computed over the model reaction coordinate for the peptide in each saline assay as well as in the absence of ions. Circular dichroism spectra were also calculated from each trajectory. In the presence of Na+ and K+ ions, no increase in helicity is observed with respect to the conformation in pure water

    Motion integration using competitive priors

    Full text link
    Psychophysical experiments show that humans are better at perceiving rotation and expansion than translation [5][9]. These findings are inconsistent with standard models of motion integration which predict best performance for translation. To explain this discrepancy, our theory formulates motion perception at two levels of inference: we first perform model selection between the competing models (e.g. translation, rotation, and expansion) and then estimate the velocity using the selected model. We define novel prior models for smooth rotation and expansion using techniques similar to those in the slow-and-smooth model [23] (e.g. Green functions of differential operators). The theory gives good agreement with the trends observed in four human experiments
    • 

    corecore